Climate change has become one of the biggest challenges of our time. Social media platforms such as Twitter play an important role in raising public awareness and spreading knowledge about the dangers of the current climate crisis. With the increasing number of campaigns and communication about climate change through social media, the information could create more awareness and reach the general public and policy makers. However, these Twitter communications lead to polarization of beliefs, opinion-dominated ideologies, and often a split into two communities of climate change deniers and believers. In this paper, we propose a framework that helps identify denier statements on Twitter and thus classifies the stance of the tweet into one of the two attitudes towards climate change (denier/believer). The sentimental aspects of Twitter data on climate change are deeply rooted in general public attitudes toward climate change. Therefore, our work focuses on learning two closely related tasks: Stance Detection and Sentiment Analysis of climate change tweets. We propose a multi-task framework that performs stance detection (primary task) and sentiment analysis (auxiliary task) simultaneously. The proposed model incorporates the feature-specific and shared-specific attention frameworks to fuse multiple features and learn the generalized features for both tasks. The experimental results show that the proposed framework increases the performance of the primary task, i.e., stance detection by benefiting from the auxiliary task, i.e., sentiment analysis compared to its uni-modal and single-task variants.
translated by 谷歌翻译
When testing conditions differ from those represented in training data, so-called out-of-distribution (OOD) inputs can mar the reliability of black-box learned components in the modern robot autonomy stack. Therefore, coping with OOD data is an important challenge on the path towards trustworthy learning-enabled open-world autonomy. In this paper, we aim to demystify the topic of OOD data and its associated challenges in the context of data-driven robotic systems, drawing connections to emerging paradigms in the ML community that study the effect of OOD data on learned models in isolation. We argue that as roboticists, we should reason about the overall system-level competence of a robot as it performs tasks in OOD conditions. We highlight key research questions around this system-level view of OOD problems to guide future research toward safe and reliable learning-enabled autonomy.
translated by 谷歌翻译
Assigning appropriate industry tag(s) to a company is a critical task in a financial institution as it impacts various financial machineries. Yet, it remains a complex task. Typically, such industry tags are to be assigned by Subject Matter Experts (SME) after evaluating company business lines against the industry definitions. It becomes even more challenging as companies continue to add new businesses and newer industry definitions are formed. Given the periodicity of the task it is reasonable to assume that an Artificial Intelligent (AI) agent could be developed to carry it out in an efficient manner. While this is an exciting prospect, the challenges appear from the need of historical patterns of such tag assignments (or Labeling). Labeling is often considered the most expensive task in Machine Learning (ML) due its dependency on SMEs and manual efforts. Therefore, often, in enterprise set up, an ML project encounters noisy and dependent labels. Such labels create technical hindrances for ML Models to produce robust tag assignments. We propose an ML pipeline which uses semantic similarity matching as an alternative to multi label text classification, while making use of a Label Similarity Matrix and a minimum labeling strategy. We demonstrate this pipeline achieves significant improvements over the noise and exhibit robust predictive capabilities.
translated by 谷歌翻译
To properly assist humans in their needs, human activity recognition (HAR) systems need the ability to fuse information from multiple modalities. Our hypothesis is that multimodal sensors, visual and non-visual tend to provide complementary information, addressing the limitations of other modalities. In this work, we propose a multi-modal framework that learns to effectively combine features from RGB Video and IMU sensors, and show its robustness for MMAct and UTD-MHAD datasets. Our model is trained in two-stage, where in the first stage, each input encoder learns to effectively extract features, and in the second stage, learns to combine these individual features. We show significant improvements of 22% and 11% compared to video only and IMU only setup on UTD-MHAD dataset, and 20% and 12% on MMAct datasets. Through extensive experimentation, we show the robustness of our model on zero shot setting, and limited annotated data setting. We further compare with state-of-the-art methods that use more input modalities and show that our method outperforms significantly on the more difficult MMact dataset, and performs comparably in UTD-MHAD dataset.
translated by 谷歌翻译
We introduce a Transformer based 6D Object Pose Estimation framework VideoPose, comprising an end-to-end attention based modelling architecture, that attends to previous frames in order to estimate accurate 6D Object Poses in videos. Our approach leverages the temporal information from a video sequence for pose refinement, along with being computationally efficient and robust. Compared to existing methods, our architecture is able to capture and reason from long-range dependencies efficiently, thus iteratively refining over video sequences. Experimental evaluation on the YCB-Video dataset shows that our approach is on par with the state-of-the-art Transformer methods, and performs significantly better relative to CNN based approaches. Further, with a speed of 33 fps, it is also more efficient and therefore applicable to a variety of applications that require real-time object pose estimation. Training code and pretrained models are available at https://github.com/ApoorvaBeedu/VideoPose
translated by 谷歌翻译
随着输入分布在任务寿命中的发展,保持基于学习的模型的性能变得具有挑战性。本文提出了一个框架,可以通过选择标签的测试输入子集来逐步重新训练模型,从而使模型适应更改输入分布。根据(1)整个任务生命周期的模型性能以及(2)与标签和模型再培训相关的累积成本,对此框架中的算法进行了评估。我们提供了卫星姿势估计模型的开源基准,该基准在空间中的卫星图像中训练并部署在新颖场景中(例如,不同的背景或不良行为的像素),在其中评估了算法,以通过在其能力上通过在其上进行高性能来维持高性能的能力。输入的子集。我们还提出了一种新颖的算法,以通过使用贝叶斯不确定性量化从输入中表征信息获得的信息增益,并选择一个子集,并选择一个子集,该子集使用批处理主动学习中的概念来最大化集体信息增益。我们表明,我们的算法在基准上的表现优于其他算法,例如,达到与100%输入标签的算法相当的性能,而仅标记了50%的输入,从而在任务寿命中产生了低成本和高性能。
translated by 谷歌翻译
脊髓损伤通常会导致四肢瘫痪的患者限制其活动能力。轮椅对于患者来说可能是一个很好的主张,但大多数人可以手动操作,也可以借助操纵杆操作的电动机。但是,这需要使用手,使其不适合四肢瘫痪的患者。另一方面,即使受到脑损伤的人,控制眼动的运动也保留了。监视眼睛中的运动可能是为轮椅生成控制信号的有用工具。本文是通过试图控制模仿轮椅的机器人来转换从眼睛转换为有意义的信号的一种方法。总体系统具有成本效益,并使用简单的图像处理和模式识别来控制机器人。开发了一种Android应用,在实际情况下,患者的援助可以使用该应用程序,以更加完善轮椅。
translated by 谷歌翻译
Code review is an integral part of any mature software development process, and identifying the best reviewer for a code change is a well accepted problem within the software engineering community. Selecting a reviewer who lacks expertise and understanding can slow development or result in more defects. To date, most reviewer recommendation systems rely primarily on historical file change and review information; those who changed or reviewed a file in the past are the best positioned to review in the future. We posit that while these approaches are able to identify and suggest qualified reviewers, they may be blind to reviewers who have the needed expertise and have simply never interacted with the changed files before. To address this, we present CORAL, a novel approach to reviewer recommendation that leverages a socio-technical graph built from the rich set of entities (developers, repositories, files, pull requests, work-items, etc.) and their relationships in modern source code management systems. We employ a graph convolutional neural network on this graph and train it on two and a half years of history on 332 repositories. We show that CORAL is able to model the manual history of reviewer selection remarkably well. Further, based on an extensive user study, we demonstrate that this approach identifies relevant and qualified reviewers who traditional reviewer recommenders miss, and that these developers desire to be included in the review process. Finally, we find that "classical" reviewer recommendation systems perform better on smaller (in terms of developers) software projects while CORAL excels on larger projects, suggesting that there is "no one model to rule them all."
translated by 谷歌翻译
本文通过实时主轴振动的表征,提出了一种白色盒子支持向量机(SVM)框架及其群体的优化。通过加速度和统计特征的时域响应,通过了过程失败(即侧面,侧面,侧面,鼻磨损,火山口和凹槽磨损,边缘骨折)而演化的异常时刻。使用作为估计器的横跨验证(RFECV)的递归特征消除,因为估计器已经用于特征选择。此外,已经检查了标准SVM的能力,用于刀具健康监测,然后通过应用群基于群的算法进行优化。已经进行了五个元启发式算法性能的比较分析(大象放牧优化,Monarch蝶优化,Harris Hawks优化,粘液模算法和飞蛾搜索算法)。考虑到全局和本地表示,已经介绍了白盒方法,这些代表可以深入了解工具状况监控中机器学习模型的性能。
translated by 谷歌翻译
我们建议承担义义歧义(WSD)的问题。在语言中,相同形式的单词可能取决于上下文。虽然人类可以通过他们的上下文轻松推断出这些单词的含义或光泽,但机器偶然地推断出这个任务。我们打算在黄等人的结果上复制和扩展他们设计消除这些词语的模型(Huang等人。,2019)。具体来说,我们提出了以下增强:数据集调整(Alpha Hyper-参数),集合方法,用BART和Albert更换BERT。以下GitHub存储库包含本报告中使用的所有代码,它延伸到Huang等人提供的代码。
translated by 谷歌翻译